Goto

Collaborating Authors

 loss predictor





A Appendix

Neural Information Processing Systems

However, these methods were ineffective in our experiment. As we explain in Section 3.2, our test-time augmentation space consists of 12 operations. Figure 4 shows a selected data sample and its augmented versions. PIL.ImageEnhance.Sharpness function gives a blurred image with parameters less The original image was distorted by some corruptions, such as rotation and noise. We use up-to 64 nodes to parallelize data-generating process.



efficient instance-aware test-time augmentation method resulting in significant gains over previous approaches

Neural Information Processing Systems

We would like to thank you for your thorough evaluation, helpful suggestions, and comments. We trained our loss predictor for five crop areas. Compared to the 5-crop ensemble, choosing one transform by our method gives almost the same performance, and selecting the two transforms achieves even better performance with less computational cost. Figure 2: Comparison for the same GPS transforms on the clean ImageNet set using ResNet-50. We trained our loss predictor on the searched GPS policies to choose ones specific for each test instance.


When does a predictor know its own loss?

Gollakota, Aravind, Gopalan, Parikshit, Karan, Aayush, Peale, Charlotte, Wieder, Udi

arXiv.org Artificial Intelligence

Given a predictor and a loss function, how well can we predict the loss that the predictor will incur on an input? This is the problem of loss prediction, a key computational task associated with uncertainty estimation for a predictor. In a classification setting, a predictor will typically predict a distribution over labels and hence have its own estimate of the loss that it will incur, given by the entropy of the predicted distribution. Should we trust this estimate? In other words, when does the predictor know what it knows and what it does not know? In this work we study the theoretical foundations of loss prediction. Our main contribution is to establish tight connections between nontrivial loss prediction and certain forms of multicalibration, a multigroup fairness notion that asks for calibrated predictions across computationally identifiable subgroups. Formally, we show that a loss predictor that is able to improve on the self-estimate of a predictor yields a witness to a failure of multicalibration, and vice versa. This has the implication that nontrivial loss prediction is in effect no easier or harder than auditing for multicalibration. We support our theoretical results with experiments that show a robust positive correlation between the multicalibration error of a predictor and the efficacy of training a loss predictor.


Learning Loss for Test-Time Augmentation

Kim, Ildoo, Kim, Younghoon, Kim, Sungwoong

arXiv.org Artificial Intelligence

Data augmentation has been actively studied for robust neural networks. Most of the recent data augmentation methods focus on augmenting datasets during the training phase. At the testing phase, simple transformations are still widely used for test-time augmentation. This paper proposes a novel instance-level test-time augmentation that efficiently selects suitable transformations for a test input. Our proposed method involves an auxiliary module to predict the loss of each possible transformation given the input. Then, the transformations having lower predicted losses are applied to the input. The network obtains the results by averaging the prediction results of augmented inputs. Experimental results on several image classification benchmarks show that the proposed instance-aware test-time augmentation improves the model's robustness against various corruptions.